Adversarial Meta Sampling for Multilingual Low-Resource Speech Recognition

نویسندگان

چکیده

Low-resource automatic speech recognition (ASR) is challenging, as the low-resource target language data cannot well train an ASR model. To solve this issue, meta-learning formulates for each source into many small tasks and meta-learns a model initialization on all from different languages to access fast adaptation unseen languages. However, languages, quantity difficulty vary greatly because of their scales diverse phonological systems, which leads task-quantity task-difficulty imbalance issues thus failure multilingual (MML-ASR). In work, we problem by developing novel adversarial meta sampling (AMS) approach improve MML-ASR. When in MML-ASR, AMS adaptively determines task probability language. Specifically, language, if query loss large, it means that its are not sampled terms should be more frequently extra learning. Inspired fact, feed historical domain network learn policy adversarially increasing current Thus, learnt can master learning situation predicts good effective Finally, experiment results two datasets show significant performance improvement when applying our also demonstrate applicability other transfer approaches.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Articulatory Feature based Multilingual MLPs for Low-Resource Speech Recognition

Large vocabulary continuous speech recognition is particularly difficult for low-resource languages. In the scenario we focus on here is that there is a very limited amount of acoustic training data in the target language, but more plentiful data in other languages. In our approach, we investigate approaches based on Automatic Speech Attribute Transcription (ASAT) framework, and train universal...

متن کامل

Multilingual Data Selection for Low Resource Speech Recognition

Feature representations extracted from deep neural networkbased multilingual frontends provide significant improvements to speech recognition systems in low resource settings. To effectively train these frontends, we introduce a data selection technique that discovers language groups from an available set of training languages. This data selection method reduces the required amount of training ...

متن کامل

Multilingual Recurrent Neural Networks with Residual Learning for Low-Resource Speech Recognition

The shared-hidden-layer multilingual deep neural network (SHL-MDNN), in which the hidden layers of feed-forward deep neural network (DNN) are shared across multiple languages while the softmax layers are language dependent, has been shown to be effective on acoustic modeling of multilingual low-resource speech recognition. In this paper, we propose that the shared-hidden-layer with Long Short-T...

متن کامل

A comparative study of BNF and DNN multilingual training on cross-lingual low-resource speech recognition

This paper presents a comparative study of bottle-neck feature (BNF) and Deep Neural Network (DNN) multilingual training for low-resource language speech recognition. Besides, we also compared the system performances after fine-tuning. The evaluation was conducted on the IARPA Babel data. The source languages are Cantonese, Pashto, Tagalog, and Turkish, while the target languages are Vietnamese...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i16.17661